RCTV.
W/18 · 2026
ISSUE 18 A publication on AI video.Written for people who already know what Sora is. APR 30 · 2026
Living Reference

AI Video Generation Tools 2026: Live Comparison & Rankings

The most current comparison of AI video generation tools in 2026 — Veo 3.1, Kling 3.0, Seedance 2.0, Runway, Luma, HappyHorse and more. Updated every week as models launch and pricing changes.

UPDATED APR 27, 2026 29 MIN READ

This is RCTV’s living reference to the AI video generation landscape. Updated as models launch, pricing changes, and capabilities evolve. Last updated: April 27, 2026.


Quick Reference: All Models at a Glance

Model Best For Max Resolution Free Tier Paid From API
Veo 3.1 (Google DeepMind) Photorealism, widest free access 4K ✓ 10 clips/mo via Google Vids $19.99/mo (AI Pro) ✓ (also via Adobe Firefly)
Kling 3.0 / 3.0 Omni (Kuaishou) Broadcast-ready 4K, 60fps, multi-shot storyboards 4K native ~$8/mo ✓ (also via Adobe Firefly)
Seedance 2.0 Pro (ByteDance) Character consistency, multi-shot 2K Via CapCut (US, with restrictions) Via CapCut / third-party Via third-party
Luma Ray 3.14 (Luma AI) Production volume, cost efficiency 1080p native Available
Runway Gen-4 Turbo / Gen-4.5 Stylized/VFX, real-time avatars 1080p $12/mo ✓ (Gen-4.5 also via Adobe Firefly)
Pika 2.5 (Pika Labs) Budget creators, rapid iteration 1080p $8/mo
Grok Imagine (xAI) Speed, cheapest API 720p (1080p Pro coming) X Premium / SuperGrok ✓ $4.20/min
LTX-2.3 (Lightricks) Local / private generation 4K ✓ Open source Free (Apache 2.0) ComfyUI
HappyHorse-1.0 (Alibaba) #1 benchmark T2V + I2V TBD Beta — weights coming soon Free (open source)
Wan 2.7 (Alibaba) Thinking Mode, 5 unified task types 1080p ✓ Open source Free ComfyUI / Model Studio
SkyReels V4 (Skywork AI) Joint audio-video, open-source 1080p ✓ 70 credits/mo Free (open source)

Rankings and pricing change weekly. Scroll down for full model breakdowns.


The Big Seven: Commercial Models

These are the production-grade models dominating professional and creator workflows in early 2026. The market has matured to the point where no single model leads across all dimensions — the professional standard is now multi-model routing, choosing the right tool for each specific shot.

Sora 2 — OpenAI

Status: Discontinued March 24, 2026; consumer app shutdown executed April 26, 2026. RCTV analysis →

OpenAI announced Sora’s discontinuation on March 24, 2026 — the app, the API, and the Disney licensing deal that was announced with it in December 2025. The stated reason was compute reallocation toward “world simulation for robotics.” The numbers tell the fuller story: estimated $15M/day peak inference cost against $2.1M in total lifetime in-app revenue, and a 66% download decline from its November 2025 peak to February 2026. Sora is removed from active tracking. See Weekly Roundup — March 27, 2026 for the full breakdown.

Shutdown timeline: The Sora consumer app and web interface went dark on April 26, 2026 (Weekly Roundup — April 27, 2026) — the export window closed at that time. The Sora API remains accessible through September 24, 2026, giving developers time to migrate integrations before the model line fully retires.


Luma Ray 3.14 — Luma AI

Best for: Professional production volume, 1080p native output, cost-efficient multi-shot workflows

Luma AI’s Ray 3.14 is the most significant release of the week — and the model that steps into the commercial tier vacated by Sora’s shutdown. See Weekly Roundup — March 27, 2026 for the full launch breakdown. Native 1080p output, generation speed 4× faster than the previous Ray 3 model, and per-second pricing 3× cheaper. Ray3 Modify, a companion tool for hybrid performance and acting workflows, gives brands and studios more control over scene continuity and character consistency across shots.

Luma is positioning Ray explicitly as professional infrastructure priced for production volume rather than a consumer app — a distinction that looks strategically deliberate given Sora’s failure. The company’s $900M Series C led by HUMAIN, new London office, and enterprise Luma Agents deployments at Publicis, Adidas, and Mazda all reinforce this direction. The Mazda relationship produced a concrete deliverable in mid-April: Boundless, a Johannesburg agency, used Luma Agents to deliver Mazda’s first AI-produced commercial in under two weeks — the most credible production-deployment signal for any AI video platform this year.

  • Max resolution: 1080p native
  • Key features: Ray3 Modify (hybrid performance/acting control), Luma Agents platform (enterprise creative automation)
  • Speed: 4× faster generation than previous Ray model
  • Pricing: 3× cheaper per-second than previous Ray
  • Access: Luma AI subscription; free tier available
  • API: Available; enterprise deployments via Luma Agents

Kling 3.0 / Kling 3.0 Omni — Kuaishou

Best for: Feature density, broadcast-ready output, motion quality

The most capability-dense model available. Kling 3.0 is the first AI video model to meet broadcast delivery standards without upscaling, offering native 4K at 60fps. The storyboard feature generates up to six camera cuts in a single generation with visual consistency — a production-first capability no other model matches. The Kling 3.0 Omni variant adds finer-grained controls for shot duration, camera angle, and character movement across multi-shot sequences.

On April 15, 2026, both Kling 3.0 and Kling 3.0 Omni were added to Adobe Firefly’s multi-model video hub alongside Veo 3.1, Runway Gen-4.5, and 30+ other AI models — significantly broadening Kling’s distribution to Adobe Creative Cloud’s existing professional user base.

  • Max resolution: 4K native (60fps)
  • Frame rate: Up to 60fps
  • Audio: Native built-in audio in six languages
  • Key feature: Multi-cut storyboard generation (up to 6 camera cuts, 15s); Omni adds shot/camera/character controls
  • Access: Free tier available; paid plans from ~$8/mo; also via Adobe Firefly (Creative Cloud subscription)
  • API: Available via Kuaishou and third-party platforms

Veo 3.1 — Google DeepMind

Best for: Photorealism, 4K native output, integrated workflows, broadest free access

Google’s model pushes photorealistic rendering to the point where trained observers struggle to identify generated footage in blind tests. The engine behind Google Flow (merged creative workspace with Whisk, ImageFX, and multi-clip sequencing) and now Google Vids, which received a major update this week.

As of April 3, 2026, any Google account holder can generate Veo 3.1 clips for free via Google Vids (Weekly Roundup — April 4, 2026) — 10 generations per month, 8 seconds at 720p, from text prompts or uploaded images. Google AI Pro and Ultra subscribers unlock more: up to 1,000 Veo clips per month, Lyria 3 custom music generation (tracks up to 3 minutes), customizable AI avatars with scene placement and wardrobe control, and direct YouTube export. This is the first time a production-grade AI video model has been made freely accessible to Google’s full account base.

On the developer side, Google launched Veo 3.1 Lite on March 31 via the Gemini API (Weekly Roundup — April 4, 2026) and Google AI Studio — priced at $0.05/sec for 720p and $0.08/sec for 1080p, less than half the cost of the existing Veo 3.1 Fast tier at the same generation speed. Supports text-to-video and image-to-video in 4-, 6-, and 8-second lengths, landscape and portrait. Veo 3.1 Fast received a further price reduction on April 7 as committed (Weekly Roundup — April 11, 2026) — compressing the full developer stack from the free consumer tier (Google Vids) through production-grade API calls. Check the Gemini API pricing documentation for current per-second rates.

  • Max resolution: 4K native (Flow/Vertex AI); 1080p via Veo 3.1 Lite; 720p via Google Vids free tier
  • Audio: Native synchronized audio
  • Key features: Flow unified workspace; Google Vids integration (avatars, Lyria 3 music, YouTube export); Veo 3.1 Lite developer tier
  • Access: Free — 10 clips/month via Google Vids (any Google account); Google AI Pro ($19.99/mo) and Ultra for higher limits; Flow is free; also via Adobe Firefly multi-model hub (April 15)
  • API: Vertex AI ($12/min); Veo 3.1 Lite via Gemini API ($0.05/sec 720p, $0.08/sec 1080p); Veo 3.1 Fast pricing reduced April 7 (check Gemini API docs for current per-second rates)
  • Milestone: 1.5 billion images and videos created by Flow users

Seedance 2.0 Pro — ByteDance

Best for: Character consistency, cinematic motion, multi-shot storytelling

The leading commercial model for character consistency and cinematic motion quality — and until this week, the top-ranked model on Artificial Analysis for both text-to-video and image-to-video. On April 10, Alibaba’s HappyHorse-1.0 displaced it to second place with an Elo score of 1,333 in T2V (Weekly Roundup — April 11, 2026) (roughly 60 points above Seedance’s position). Seedance 2.0 Pro’s Dual-Branch Diffusion Transformer generates audio and video simultaneously in a single pass. Its quad-modal input system accepts text, images, video, and audio in a single prompt. Multi-shot native storytelling and frame-level control over character appearance, object placement, and scene timing remain best-in-class for narrative work.

ByteDance’s official global API rollout was paused indefinitely in late February 2026 after the Motion Picture Association and major studios (Disney, Netflix, Paramount, Sony, Warner Bros.) issued cease-and-desist letters over copyright concerns. The “Face-to-Voice” feature was suspended on February 10 after it was shown to clone voices from a single photo. Japan opened a separate inquiry over unauthorized anime character reproductions.

On March 23, 2026, ByteDance moved anyway — relaunching the model as Dreamina Seedance 2.0 (Weekly Roundup — March 27, 2026) across markets in Africa, South America, the Middle East, and Southeast Asia. As of April 2026, Dreamina Seedance 2.0 is now rolling out in the US via CapCut (Weekly Roundup — April 11, 2026) — a significant reversal of the prior exclusion. The deployment comes with content restrictions: image-to-video generation from inputs containing real faces is disabled, and generation of unauthorized intellectual property is blocked. All output carries an invisible watermark for off-platform identification. The rollout is phased and may not have reached all US CapCut accounts simultaneously.

The copyright landscape around Seedance grew more complex this week. The White House published its National Policy Framework for AI (March 2026) stating that AI training on copyrighted works does not constitute infringement — the opposite of the Blackburn bill’s position. Separately, the bipartisan CLEAR Act (Schiff/Curtis) would require public disclosure of training data without resolving the fair use question either way. Three incompatible federal positions now coexist: Blackburn (training = not fair use), White House (training = fine, let courts decide), CLEAR Act (disclose and defer).

  • Max resolution: 2K
  • Audio: Native audio with lip-sync
  • Key feature: Multi-shot storytelling, quad-modal input, frame-level precision
  • Access: China via Jimeng/Dreamina; Africa, South America, Middle East, SE Asia, and now US via CapCut/Dreamina Seedance 2.0; global via BigMotion ($35–$95/mo), LumeFlow AI, other third-party platforms
  • API: Official global API paused; available via third-party integrations (fal.ai, others)
  • US restrictions: Real-face image-to-video disabled; unauthorized IP generation blocked; invisible watermarks on all output
  • Note: Benchmark position: moved to #2 on Artificial Analysis (T2V) as of April 10 — displaced by Alibaba HappyHorse-1.0; copyright legislative battle remains a three-way standoff (Blackburn vs. White House vs. CLEAR Act)

Grok Imagine — xAI

Best for: Speed, low-cost API, rapid iteration, social media distribution

The newest entrant to the AI video generation field, and the fastest-iterating. xAI shipped four major updates in five weeks: API launch (January 28), Grok Imagine 1.0 with 720p video and audio (February 3), Grok 4.20 (February 17), and video extension (March 2). The “Extend from Frame” feature lets users chain clips by continuing from the final frame, enabling sequences up to 30 seconds while preserving lighting, motion, and character positioning.

On April 17, 2026, xAI released Grok 4.3 Beta with native video understanding — letting Grok analyze video as a coherent temporal sequence rather than as isolated frames. The understanding capability is distinct from Grok Imagine’s generation pipeline, but the two now stack: Grok can both generate and reason about video within the same model family. No other major lab currently offers vertical integration of native generation, native understanding, and platform-scale distribution under a single subscription.

Grok Imagine’s API pricing dramatically undercuts the field. The trade-off is a 720p resolution ceiling — every other major model offers 1080p or higher. Community testing also confirms visible quality degradation after two or three chained extensions. The distribution advantage is unique: over 500 million X users have direct access. Video features are currently locked behind X Premium subscriptions.

  • Max resolution: 720p (1080p via Grok Imagine Pro — confirmed for late April 2026)
  • Max duration: 30 seconds (via chained extensions)
  • Audio: Synchronized audio
  • Key feature: Video extension from frame; dual generation modes (Quality + Speed); native video understanding (Grok 4.3 Beta); fastest iteration cycle in the industry
  • Access: X Premium / SuperGrok subscription required
  • API: Available ($4.20/min generated video — cheapest major model)
  • Engine: Aurora autoregressive MoE model on 110,000 NVIDIA GB200 GPUs
  • Next: Grok Imagine Pro (1080p) confirmed by Elon Musk for late April 2026 — available to SuperGrok subscribers
  • Caution: Faced regulatory scrutiny over content moderation (UK ICO, France, California AG); image editing now restricted to paid subscribers

Runway Gen-4 Turbo — Runway

Best for: Stylized content, VFX aesthetics, professional ecosystem, real-time avatars

Runway leads in non-photorealistic and stylized video — VFX-oriented aesthetics, abstract content, and artistic directions where other models default to photorealism. Gen-4 Turbo also has the most mature professional ecosystem with motion brushes, scene consistency tools, and a robust API. Runway closed a $315M Series C in February 2026 at a $5.3B valuation.

On March 9, Runway launched Characters — a real-time video agent API built on its GWM-1 world model. (Weekly Roundup — March 13, 2026) Characters generates fully conversational AI avatars from a single reference image with no fine-tuning required. The avatars sustain realistic lip-sync, facial expressions, eye contact, and gesture across extended multi-minute conversations, running at 24fps at 720p in real time. BBC and Silverside are early enterprise partners, with the API available to developers at dev.runwayml.com. Characters targets customer support, interactive training, and branded marketing experiences — a direct play for the enterprise avatar market currently led by HeyGen and Synthesia.

At NVIDIA GTC (March 17, 2026), Runway demoed a research preview of a new real-time video generation model (Weekly Roundup — March 20, 2026) running on NVIDIA Vera Rubin hardware — achieving time-to-first-frame under 100ms for HD video. The demo also revealed that Gen-4.5, Runway’s next commercial model, is being developed and tested on Vera Rubin ahead of its public release. As of April 15, 2026, Gen-4.5 is accessible to creators via Adobe Firefly’s multi-model video hub — Runway’s first major distribution beyond its own platform — even though Runway has yet to formally announce a standalone Gen-4.5 launch.

  • Max resolution: 1080p (Gen-4 Turbo); 720p real-time (Characters)
  • Audio: Supported
  • Key features: Motion brushes, style control, API maturity; Characters real-time avatar API (GWM-1)
  • Access: From $12/mo (runwayml.com); Gen-4.5 also via Adobe Firefly (Creative Cloud subscription)
  • API: Most mature video generation API available; Characters API at dev.runwayml.com
  • Next: Gen-4.5 in market via Firefly (April 15); real-time video model research preview on Vera Rubin hardware (sub-100ms TTF); standalone Gen-4.5 launch on runwayml.com still pending
  • Note: Characters is an enterprise API product built on GWM-1, separate from the Gen-4 Turbo generation pipeline

Pika 2.5 — Pika Labs

Best for: Budget-conscious creators, rapid iteration, social media content

The most accessible entry point to AI video generation. Pika’s strength is speed and volume — generate 20-30 variations of a concept in minutes, then refine. Features like Pikaswaps (face/object replacement) and Pikaffects (style transfer) add creative flexibility at a price point that undercuts every competitor.

On April 2, 2026, Pika launched PikaStream 1.0 — a real-time AI video engine for live agent meetings at 24fps/480p with ~1.5s speech-to-video latency and persistent identity across calls. PikaStream is a separate product line from Pika 2.5 (live agent video, not clip generation) and signals a pivot toward AI agent infrastructure. Note: the official Pika blog has been quiet since the PikaStream launch — a notable gap for one of 2024–2025’s loudest brands.

  • Max resolution: 1080p (Pika 2.5); 480p real-time (PikaStream 1.0)
  • Max duration: 42 seconds (clip); persistent for live (PikaStream)
  • Audio: Supported
  • Key feature: Pikaswaps, Pikaffects, fast batch generation; PikaStream 1.0 for live agent video
  • Access: From $8/mo (lowest entry price among major models)
  • API: Available

Open-Source & Local Generation

The open-source AI video ecosystem has matured significantly, making local generation on consumer hardware a viable option for privacy-conscious creators and developers.

LTX-2.3 — Lightricks

Best for: Local/desktop generation, consumer GPU workflows, high-frame-rate output

The standout for local generation — and now significantly upgraded. Released March 2026 (Weekly Roundup — March 20, 2026), LTX-2.3 is a comprehensive rebuild of the original LTX-2: a new VAE for sharper detail, a 4× larger text connector for better prompt understanding, and an improved HiFi-GAN vocoder for cleaner native audio. The model ships alongside a dedicated desktop video editor, making the entire local pipeline accessible without a ComfyUI node graph.

Key new capabilities: native portrait mode (9:16 up to 1080×1920), last-frame interpolation for seamless clip chaining, and 24/48fps output options alongside the previous frame rates. At GDC this week, NVIDIA announced 2.5× performance gains on RTX 50 Series via NVFP4 quantization, 60% lower VRAM usage, and RTX Video Super Resolution for ComfyUI delivering 4K upscaling 30× faster than competing local alternatives. The new ComfyUI App View strips the node-graph interface into a simplified prompt-in/video-out UI for non-technical users.

  • Max resolution: 4K native (true 4K, not upscaled)
  • Max duration: 20 seconds
  • Frame rate: Up to 50fps (24/48fps options also available)
  • Audio: Native synchronized audio (improved HiFi-GAN vocoder)
  • Portrait mode: Yes (9:16, up to 1080×1920)
  • Hardware: Runs on GPUs with 12GB+ VRAM; optimized for RTX 50 Series (2.5× faster via NVFP4)
  • Integration: ComfyUI native; standalone desktop video editor (shipped March 2026)
  • License: Apache 2.0 (free for companies under $10M revenue; commercial license required above that threshold)

Wan 2.7 — Alibaba (Tongyi Lab)

Best for: Multi-task video generation with Thinking Mode, open-source flexibility

Alibaba’s Wan 2.7, released April 3, 2026 (Weekly Roundup — April 17, 2026), is a major upgrade from the 2.2 line. The headline feature is Thinking Mode — a chain-of-thought reasoning approach where the model analyzes the prompt, plans composition, then generates. This produces noticeably more coherent output with fewer artifacts than single-pass generation.

Wan 2.7 Video unifies five task types in a single model: text-to-video, image-to-video (first-frame, first-and-last-frame, audio-driven), video continuation with text guidance, reference-to-video with up to five real-person inputs, and video editing via text, reference images, or style transfer. ComfyUI added support the same day in version 0.18.5 with workflow templates for all five task types.

  • Max resolution: 1080p
  • Max duration: 2–15 seconds
  • Task types: T2V, I2V, video continuation, reference-to-video (up to 5 persons), video editing
  • Key feature: Thinking Mode (chain-of-thought reasoning before generation)
  • Integration: ComfyUI 0.18.5+, Alibaba Cloud Model Studio, wan.video
  • License: Open source

Other Notable Open-Source Models

  • HappyHorse-1.0 (Alibaba ATH AI Innovation Unit) — Debuted anonymously on Artificial Analysis on April 7, 2026 (Weekly Roundup — April 11, 2026), and immediately ranked #1 in both text-to-video and image-to-video blind testing, surpassing Seedance 2.0. Alibaba revealed its ownership on April 10. As of April 17, Elo scores have risen further: T2V (no audio) at 1,364, I2V (no audio) at 1,398. The 15B-parameter model uses a unified 40-layer self-attention Transformer processing text, image, video, and audio tokens in the same sequence. Still in beta; GitHub and model weights listed as “coming soon.” ATH also released Happy Oyster on April 16 (Weekly Roundup — April 17, 2026) — a world model that generates interactive, physics-aware 3D environments from text prompts, targeting gaming, film, and VR. Live demo accessible via Artificial Analysis arena.
  • SkyReels V4 (Skywork AI) — Released April 3, 2026 (Weekly Roundup — April 17, 2026). First open-source model to co-generate video and synchronized audio in a single forward pass. Dual-stream Multimodal Diffusion Transformer (MMDiT) architecture; 1080p at 32 FPS, clips up to 15 seconds. Accepts text, images, video clips, masks, and audio references. Ranked among the top models on Artificial Analysis T2V with audio leaderboard (Elo ~1,135). Free tier: 70 monthly credits on skyreels.dev; open-source weights available for local deployment
  • Mochi 1 — High-fidelity short video with strong prompt alignment
  • HunyuanVideo / HY-World 2.0 (Tencent) — HunyuanVideo offers solid image-to-video with coherent motion. On April 16, 2026, Tencent’s Hunyuan team released HY-World 2.0 — a multi-modal world model that generates editable 3D scenes (meshes plus Gaussian Splattings) from text prompts or single reference images, with WorldMirror 2.0 inference code and weights open-sourced (github.com/Tencent-Hunyuan/HY-World-2.0). The combination of editable 3D geometry and open weights makes HY-World 2.0 the more pipeline-friendly counterpart to Alibaba’s still-gated Happy Oyster
  • MAGI-1 — Long-form video synthesis capabilities
  • Helios (Peking University / ByteDance / Canva) — 14B autoregressive diffusion model; 19.5fps real-time generation on a single NVIDIA H100; capable of minute-scale video; Apache 2.0 license. Released March 2026. Notable for real-time throughput on a single accelerator.

How to Choose: A Routing Framework

The right model depends on the shot, not the project. Here’s a practical decision framework:

Need broadcast-ready 4K? → Kling 3.0 or Veo 3.1

Need a free starting point? → Veo 3.1 via Google Vids (10 free clips/month, any Google account)

Need character consistency across shots? → Seedance 2.0 Pro via CapCut (US now available, with real-face restrictions) or Luma Ray 3.14

Need stylized / VFX aesthetic? → Runway Gen-4 Turbo

Need professional production volume at scale? → Luma Ray 3.14 (4× faster, 3× cheaper than previous Ray)

Need volume at low cost? → Pika 2.5

Need cheapest API? → Grok Imagine ($4.20/min)

Need local generation / privacy? → LTX-2.3 via ComfyUI or desktop editor

Need real-time interactive avatars? → Runway Characters (GWM-1)

Need real-time live AI video for agent meetings? → PikaStream 1.0 (24fps/480p, ~1.5s latency)

Need multi-shot narrative? → Seedance 2.0 Pro via CapCut (US, with restrictions), Luma Ray 3.14, or Kling 3.0 Omni

Need Adobe Creative Cloud integration? → Adobe Firefly multi-model hub (Veo 3.1, Kling 3.0/Omni, Runway Gen-4.5, Luma, plus 30+ others)

Need editable 3D world / scene generation? → Tencent HY-World 2.0 (open weights) or Alibaba Happy Oyster (gated early access)

Need massive distribution? → Grok Imagine (500M+ X users)

Most professional workflows now use 2-3 models per project, routing different shots to different engines based on the specific requirements of each scene.


What’s Coming

  • Grok Imagine Pro (1080p) — Confirmed by Elon Musk for late April 2026. Upgrades resolution from 720p to 1080p for SuperGrok subscribers. If delivered at the existing $4.20/min API rate, this significantly changes Grok Imagine’s competitive positioning
  • HappyHorse-1.0 open-source weights — Alibaba’s ATH AI Innovation Unit confirmed the model will be fully open-sourced; GitHub and model weights listed as “coming soon” as of April 10. When released, HappyHorse-1.0 will be the new benchmark leader for open-source video generation by a significant margin
  • TAKE IT DOWN Act platform compliance — May 19, 2026 deadline for covered platforms to comply with notice-and-takedown obligations for intimate visual depictions and deepfakes. The April 8 first criminal conviction raises the legal profile of this deadline. Expect platform policy announcements in the run-up
  • Runway Gen-4.5 — Now accessible via Adobe Firefly’s multi-model hub (April 15, 2026); standalone Gen-4.5 launch on runwayml.com still pending. Previewed on NVIDIA Vera Rubin hardware at GTC (March 2026); expected to be Runway’s highest-performance commercial model to date
  • NVIDIA Vera Rubin cloud deployment — AWS, Google Cloud, Microsoft Azure, and OCI all confirmed H2 2026 availability. Vera Rubin delivers 10× lower inference token cost versus Blackwell — the number that will reshape per-second AI video pricing across all major cloud platforms
  • DLSS 5 — NVIDIA’s neural rendering technology, launching Fall 2026. Explicitly positioned for filmmaking and VFX beyond gaming; uses generative AI to infuse photoreal lighting and materials anchored to source 3D geometry
  • Blackburn draft AI bill — GOP Senate draft (March 19, 2026) declares AI training on copyrighted works not fair use; targets deepfakes and Section 230. Not yet introduced as legislation; path to passage uncertain
  • White House AI framework vs. CLEAR Act — White House (March 2026) takes the opposite position from Blackburn: AI training is not infringement; courts should decide. Bipartisan CLEAR Act (Schiff/Curtis) proposes mandatory training data disclosure without resolving fair use. Three irreconcilable positions now active in Washington simultaneously
  • Seedance 2.0 copyright litigation — US CapCut access now live (April 2026) with real-face and IP restrictions, but the underlying copyright dispute with Disney, Paramount, Warner Bros., and Netflix remains unresolved. The restrictions are a negotiating posture, not a settlement
  • OpenAI robotics / world simulation — OpenAI redirected Sora’s compute toward “world simulation for robotics” after shutting the product down. The consumer app went dark on April 26, 2026 as scheduled; the Sora API remains accessible until September 24, 2026. Watch for whether any video-adjacent capability resurfaces inside ChatGPT before that deadline
  • Adobe Firefly multi-model expansion — Firefly’s video hub now hosts 30+ third-party AI models including Kling 3.0/Omni, Veo 3.1, Runway Gen-4.5, ElevenLabs Multilingual v2, Luma AI, Black Forest Labs, and Topaz Labs. Firefly AI Assistant orchestrates multi-step workflows across Photoshop, Premiere, Lightroom, Express, and Illustrator. Watch for additional model integrations and any model labs that opt to remain Firefly-exclusive
  • Tencent vs. Alibaba 3D world model race — Two of China’s largest AI labs shipped 3D world models on the same day, April 16 (Alibaba’s Happy Oyster, gated; Tencent’s HY-World 2.0, open weights). Western labs have nothing comparable in production; the 6-to-12 month head start is real if world simulation matters as much as OpenAI’s Sora-shutdown framing implied
  • Google Vids / Workspace expansion — YouTube export is now live; paid creative tiers (Pro/Ultra) now include Lyria 3 music generation and AI avatars. Further Workspace AI integration expected throughout 2026
  • EU AI Act Article 50 — August 2026, requiring machine-readable metadata on all AI-generated content
  • Unlimited-length AI video — EPFL’s drift elimination breakthrough (presenting at ICLR 2026) could remove the duration ceiling entirely
  • xAI targeting 30-minute video — Announced goal for late 2026, with full-length films targeted for 2027

This page is maintained by RCTV as a public reference. For weekly updates on model releases and industry shifts, see our Weekly Roundup.

Have a correction or update? Contact us at rctv.oxncw@simplelogin.com


CHANGELOG

  • Last updated date: Advanced from March 7 to March 13, 2026
  • Sora 2 (OpenAI): Added Sora 1 retirement notice (effective March 13, US users); added note on reported ChatGPT integration plans (The Information, March 12) and declining standalone app install figures (−45% MoM, January 2026); updated Disney licensing detail to specify all four brands (Disney, Marvel, Pixar, Star Wars); added Note spec bullet summarizing access changes
  • Runway Gen-4 Turbo: Added $315M Series C / $5.3B valuation (February 2026); added two-paragraph description of Runway Characters (launched March 9) — real-time avatar API built on GWM-1, BBC and Silverside as launch partners, available at dev.runwayml.com; updated Best for tagline and spec bullets to reflect Characters alongside Gen-4 Turbo
  • LTX-2 → LTX-2.3: Renamed entry to reflect new version; rewrote description to cover the 22B-parameter rebuild (new VAE, 4× text connector, HiFi-GAN vocoder), desktop video editor launch, portrait mode (9:16), last-frame interpolation, and 24/48fps options; updated NVIDIA acceleration figures from “3× / 60%” (CES) to “2.5× / 60%” (GDC RTX 50 Series specific); added RTX Video Super Resolution 30× upscaling stat; added ComfyUI App View simplified UI; updated spec bullets (true 4K native, 50fps, Apache 2.0 commercial licensing threshold)
  • Other Notable Open-Source Models: Added Helios (Peking University / ByteDance / Canva) — 14B autoregressive diffusion model, 19.5fps real-time on single H100, Apache 2.0, released March 2026
  • How to Choose routing framework: Updated “LTX-2” reference to “LTX-2.3”; added “Need real-time interactive avatars? → Runway Characters (GWM-1)”
  • What’s Coming: Added “Sora integration into ChatGPT” item; updated “NVIDIA GTC 2026” from “later this month” to specific March 16–19 date with Jensen Huang keynote detail; all other items unchanged

March 20, 2026

  • Last updated date: Advanced from March 13 to March 20, 2026
  • Tags: Added regulation tag — regulatory developments are now a recurring weekly topic affecting multiple tracked models
  • Runway Gen-4 Turbo: Added third description paragraph covering the GTC real-time video model research preview (sub-100ms TTF on Vera Rubin hardware, March 17) and Gen-4.5 signal (confirmed running on Vera Rubin ahead of public release, no launch date); added Next: spec bullet summarizing both developments
  • Seedance 2.0 Pro: Added paragraph noting legislative reinforcement from the Blackburn draft AI bill (March 19) — which, if enacted, would make unauthorized AI training use of copyrighted works categorically not fair use; updated Note: spec bullet to reference the bill
  • What’s Coming: Removed “NVIDIA GTC 2026” item (event concluded); added “Runway Gen-4.5” (previewed at GTC); added “NVIDIA Vera Rubin cloud deployment H2 2026” (confirmed by AWS, Google Cloud, Microsoft, OCI; 10× inference cost reduction); added “DLSS 5 Fall 2026” (neural rendering with VFX/filmmaking ambitions beyond gaming); added “Blackburn draft AI bill” (copyright/deepfake provisions); updated Seedance item to note Blackburn bill adds legislative pressure

March 27, 2026

  • Last updated date: Advanced from March 20 to March 27, 2026
  • Tags: Added luma tag to reflect Luma Ray 3.14 promotion to primary tracked model
  • Sora 2 — OpenAI: Removed from active Big Seven tracking — product discontinued March 24, 2026. Section replaced with a brief discontinuation note citing $15M/day burn rate vs. $2.1M lifetime revenue, Disney deal collapse, and OpenAI’s pivot to robotics/world simulation. Model entry retained as a historical reference with strikethrough heading
  • Luma Ray 3.14 — Luma AI: Added as new Big Seven entry, replacing Sora. Native 1080p, 4× faster generation, 3× cheaper per-second pricing vs. previous Ray. Ray3 Modify companion tool for hybrid performance/acting workflows. $900M Series C (HUMAIN), new London office. Enterprise deployments via Luma Agents (Publicis, Adidas, Mazda). Promoted from “Models to Watch” list
  • Seedance 2.0 Pro: Added paragraph on Dreamina Seedance 2.0 relaunch (March 23) in Africa, South America, Middle East, Southeast Asia — US still excluded. Added paragraph on three-way Washington copyright standoff: Blackburn bill vs. White House AI framework vs. CLEAR Act (Schiff/Curtis). Updated Access and Note spec bullets to reflect Dreamina regional launch
  • How to Choose routing framework: Removed “Need realistic physics? → Sora 2”; added “Need professional production volume at scale? → Luma Ray 3.14”; updated “character consistency” and “multi-shot narrative” routes to include Luma Ray 3.14 alongside Seedance as a US-accessible alternative
  • What’s Coming: Removed “Sora integration into ChatGPT” (product dead); added “White House AI framework vs. CLEAR Act” (new copyright legislative developments); updated Seedance item to reflect Dreamina emerging-markets launch and US exclusion; added “OpenAI robotics/world simulation” item tracking what comes next from OpenAI in the video-adjacent space

April 4, 2026

  • Last updated date: Advanced from March 27 to April 4, 2026
  • Sora 2 — OpenAI (discontinued): Added specific shutdown timeline — app and web interface off April 26, 2026; API access continues until September 24, 2026
  • Veo 3.1 — Google DeepMind: Major access and pricing update. Added Google Vids free tier (10 clips/month for any Google account, 8 sec at 720p, text-to-video and image-to-video); added Google AI Pro/Ultra features (1,000 clips/month, Lyria 3 music up to 3 min, AI avatars with scene/wardrobe control, YouTube export now live). Added Veo 3.1 Lite developer tier via Gemini API / Google AI Studio ($0.05/sec 720p, $0.08/sec 1080p, <50% cost of Fast tier, same speed). Updated “Best for” tagline, description paragraphs, Access spec, and API spec bullets. Updated Key features bullet to reflect Vids integration
  • How to Choose routing framework: Added “Need a free starting point? → Veo 3.1 via Google Vids (10 free clips/month)”
  • What’s Coming: Removed “Google Flow + YouTube integration” (YouTube export now live in Google Vids); added “Veo 3.1 Fast price cut April 7”; updated “OpenAI robotics/world simulation” item to include Sora API September 24 deadline; replaced with “Google Vids/Workspace expansion” item for ongoing paid tier rollout

April 11, 2026

  • Last updated date: Advanced from April 4 to April 11, 2026
  • Seedance 2.0 Pro: Updated benchmark position — no longer top-ranked on Artificial Analysis; displaced to #2 T2V by HappyHorse-1.0 (Elo 1,333) as of April 10. Updated US access status: Dreamina Seedance 2.0 now rolling out in US CapCut with real-face image-to-video disabled, unauthorized IP blocked, and invisible watermarking on all output. Updated Access, API, and Note spec bullets accordingly. Updated description paragraph to reflect that US was excluded but is now partially available
  • Veo 3.1 — Google DeepMind: Confirmed Veo 3.1 Fast price reduction occurred April 7 as committed; removed “scheduled for April 7” language, updated API bullet to direct developers to Gemini API docs for current per-second rates; updated description paragraph to reflect the cut landed
  • Other Notable Open-Source Models: Added HappyHorse-1.0 (Alibaba ATH AI Innovation Unit) — #1 T2V at Elo 1,333, #1 I2V at Elo 1,392 on Artificial Analysis; model in beta, weights and GitHub “coming soon”; expected to become the top open-source video model on release
  • How to Choose routing framework: Updated Seedance routing entries to remove “(if accessible)” and add “via CapCut (US, with restrictions)” to reflect partial US access
  • What’s Coming: Removed “Seedance 2.0 Pro US access” (resolved partially via CapCut with restrictions); removed “Veo 3.1 Fast price cut April 7” (confirmed delivered); added “HappyHorse-1.0 open-source weights” (coming soon per ATH team); added “TAKE IT DOWN Act platform compliance — May 19, 2026”; updated “Seedance copyright litigation” item to reflect current state with US CapCut access but unresolved studio disputes

April 17, 2026

  • Last updated date: Advanced from April 11 to April 17, 2026
  • Tags: Added skyreels tag to reflect SkyReels V4 promotion to tracked model
  • Quick Reference table: Updated Wan 2.2 → Wan 2.7 with new capabilities (Thinking Mode, 5 task types, ComfyUI/Model Studio access); added SkyReels V4 row (1080p, 70 free credits/mo, open source); updated Grok Imagine resolution to note “1080p Pro coming”
  • Wan 2.2 → Wan 2.7: Complete rewrite of model entry. Wan 2.7 released April 3 with Thinking Mode (chain-of-thought reasoning), 5 unified task types (T2V, I2V, video continuation, reference-to-video, video editing), ComfyUI 0.18.5 support, 1080p at 2–15 seconds
  • SkyReels V1 → SkyReels V4: Updated from V1 to V4 in Other Notable Open-Source Models. SkyReels V4 (April 3) is the first open-source model with joint audio-video generation via dual-stream MMDiT; 1080p/32FPS, up to 15 seconds; Elo ~1,135 on Artificial Analysis T2V with audio; 70 free monthly credits
  • HappyHorse-1.0: Updated Elo scores (T2V no audio: 1,361, up from 1,333; I2V no audio: 1,398, up from 1,392); added 15B parameter count and 40-layer unified Transformer architecture detail; added Happy Oyster world model release (April 16) — ATH’s interactive 3D environment generator for gaming, film, and VR
  • Grok Imagine: Updated resolution spec to note 1080p Pro confirmed for late April 2026; added Quality + Speed dual generation modes (early April); updated engine description to “Aurora autoregressive MoE model”; added Next bullet for Pro tier; updated Access to include SuperGrok
  • What’s Coming: Added “Grok Imagine Pro (1080p)” item — confirmed by Musk for late April, available to SuperGrok subscribers

April 20, 2026

  • HappyHorse-1.0: T2V (no audio) Elo updated to 1,364 (from 1,361 as of April 17)

April 27, 2026

  • Last updated date: Advanced from April 17 to April 27, 2026
  • Tags: Added adobe, firefly, hunyuan, world-models — Adobe Firefly’s multi-model hub is now a primary distribution surface for tracked models; Tencent HY-World 2.0 establishes Hunyuan as a separate open-source thread; world-models is now a recurring beat
  • Quick Reference table: Updated Veo 3.1, Kling 3.0, and Runway Gen-4 Turbo rows to flag Adobe Firefly as an additional access channel; renamed Kling row to “Kling 3.0 / 3.0 Omni” to surface the new variant; renamed Runway row to “Runway Gen-4 Turbo / Gen-4.5” to reflect Gen-4.5’s appearance via Firefly
  • Sora 2 — OpenAI: Updated status line and shutdown timeline from future-tense to past-tense — consumer app and web interface went dark on April 26, 2026, as scheduled. API access still continues through September 24, 2026. Linked to April 27 weekly roundup
  • Luma Ray 3.14: Added Boundless / Mazda commercial detail — Johannesburg agency delivered Mazda’s first AI-produced commercial in under two weeks using Luma Agents (April 2026); positioned as the most credible production-deployment signal for any AI video platform this year
  • Kling 3.0 → Kling 3.0 / Kling 3.0 Omni: Renamed entry; added Kling 3.0 Omni variant with shot/camera/character controls; added paragraph on April 15 Adobe Firefly hub addition; updated Access spec to include Adobe Firefly distribution
  • Veo 3.1 — Google DeepMind: Updated Access spec to note Adobe Firefly multi-model hub availability (April 15)
  • Runway Gen-4 Turbo: Updated description to note Gen-4.5 is now accessible via Adobe Firefly (April 15) — Runway’s first major distribution beyond its own platform; updated Access, Next bullets, and What’s Coming entry accordingly
  • Grok Imagine — xAI: Added paragraph on Grok 4.3 Beta native video understanding (April 17); paired generation + understanding under one model family; updated Key feature bullet
  • Pika 2.5 — Pika Labs: Added PikaStream 1.0 (April 2) — real-time AI video engine for live agent meetings (24fps/480p, ~1.5s latency); flagged Pika blog quiet period since launch; updated spec bullets to reflect both Pika 2.5 and PikaStream
  • Other Notable Open-Source Models: Renamed HunyuanVideo entry to “HunyuanVideo / HY-World 2.0” — added Tencent’s April 16 multi-modal world model release with editable 3D scenes (meshes + Gaussian Splattings) and open weights; positioned as the more pipeline-friendly counterpart to Alibaba’s gated Happy Oyster
  • How to Choose routing framework: Added “Need real-time live AI video for agent meetings?” → PikaStream 1.0; added “Need Adobe Creative Cloud integration?” → Firefly multi-model hub; added “Need editable 3D world / scene generation?” → HY-World 2.0 or Happy Oyster; expanded multi-shot routing to include Kling 3.0 Omni
  • What’s Coming: Updated Runway Gen-4.5 entry — now in market via Firefly, standalone runwayml.com launch still pending; updated OpenAI / Sora item to reflect April 26 app shutdown executed and remaining September 24 API sunset; added “Adobe Firefly multi-model expansion” item; added “Tencent vs. Alibaba 3D world model race” item
Mondays, before 9am PT
The week in AI video, without the hype tax.